error_code &= 3;
error_code |= (regs->xcs & 2) << 1;
- /* ensure all updates have completed */
- flush_page_update_queue();
-
/*
* We fault-in kernel-space virtual memory on-demand. The
* 'reference' page table is init_mm.pgd.
}
pte_ofs = 0;
}
- flush_page_update_queue();
}
pmd_idx = 0;
}
queue_pgd_pin(__pa(new_pgd));
load_cr3(new_pgd);
queue_pgd_unpin(__pa(old_pgd));
- __flush_tlb_all(); /* implicit flush */
make_page_writable(old_pgd);
- flush_page_update_queue();
+ __flush_tlb_all(); /* implicit flush */
free_bootmem(__pa(old_pgd), PAGE_SIZE);
kernel_physical_mapping_init(new_pgd);
zone_sizes_init();
/* Switch to the real shared_info page, and clear the dummy page. */
- flush_page_update_queue();
set_fixmap_ma(FIX_SHARED_INFO, xen_start_info.shared_info);
HYPERVISOR_shared_info = (shared_info_t *)fix_to_virt(FIX_SHARED_INFO);
memset(empty_zero_page, 0, sizeof(empty_zero_page));
rx_mcl[i].args[2] = 0;
}
- /*
- * We may have allocated buffers which have entries outstanding in the page
- * update queue -- make sure we flush those first!
- */
- flush_page_update_queue();
-
/* After all PTEs have been zapped we blow away stale TLB entries. */
rx_mcl[i-1].args[2] = UVMF_FLUSH_TLB;